CS 229 Supplemental Lecture notes

نویسنده

  • John Duchi
چکیده

Building off of our interpretations of supervised learning as (1) choosing a representation for our problem, (2) choosing a loss function, and (3) minimizing the loss, let us consider a slightly more general formulation for supervised learning. In the supervised learning settings we have considered thus far, we have input data x ∈ R and targets y from a space Y . In linear regression, this corresponded to y ∈ R, that is, Y = R, for logistic regression and other binary classification problems, we had y ∈ Y = {−1, 1}, and for multiclass classification we had y ∈ Y = {1, 2, . . . , k} for some number k of classes. For each of these problems, we made predictions based on θx for some vector θ, and we constructed a loss function L : R×Y → R, where L(θx, y) measures the loss we suffer for predicting θx. For logistic regression, we use the logistic loss L(z, y) = log(1 + e) or L(θx, y) = log(1 + e T ).

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

BU CAS CS 332, Spring 2009: Section Notes

This document contains notes corresponding to the material that has been or will be covered during the discussion sections of the spring 2009 iteration of the course BU CAS CS 332, taught by Professor Leonid Levin. These notes will contain intuitive exposition, as well as problems and examples that demonstrate how the definitions and theorems presented during lectures and in the official lectur...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2016